1. The Joint Cumulative Distribution Function (JCDF)
The foundation of multi-variable analysis is the Joint Distribution Function $F(a_1, a_2, \dots, a_n)$. It defines the probability that multiple conditions are met at the same time.
$F(a_1, a_2, \dots, a_n) = P\{X_1 \le a_1, X_2 \le a_2, \dots, X_n \le a_n\}$
This formula represents the probability that each variable $X_i$ falls below its respective threshold $a_i$ simultaneously. Geometrically, in two dimensions, this is the probability that the random pair $(X, Y)$ falls within the semi-infinite rectangle to the lower-left of the point $(a, b)$.
2. The Infinitesimal Interpretation of Density
For continuous variables, we describe probability through a Joint Probability Density Function (JPDF), $f(x, y)$. Unlike discrete cases, the probability at a single point is zero. Instead, we look at infinitesimal regions:
- The probability that a pair $(X, Y)$ falls within a tiny rectangle is given by:
$P\{a < X < a + da, b < Y < b + db\} = \int_{b}^{b+db} \int_{a}^{a+da} f(x, y) \, dx \, dy \approx f(a, b) \, da \, db$ - Alternatively expressed as: $P\{x < X < x + dx, y < Y < y + dy\} \approx f(x, y) dx dy$
This reveals that $f(x, y)$ is a "density" relative to the area of the region in the Cartesian plane.
3. Dependency and Geometric Constraints
In probability, Random variables that are not independent are said to be dependent. This is not just an algebraic property; it is often visible in the support of the distribution.
Consider a point $(X, Y)$ chosen uniformly within a circle of radius $R$ centered at $(0,0)$. The variables $X$ and $Y$ are dependent because knowing $X = x$ limits the possible values of $Y$.
If $X$ is near $R$, $Y$ must be near $0$. Mathematically, $Y$ is constrained: $-\sqrt{R^2 - X^2} \le Y \le \sqrt{R^2 - X^2}$. This boundary is what prevents the joint density from being factored into independent marginals.